More or Less True DCTL for Continuous-Time MDPs
نویسنده
چکیده
Discounted Computation Tree Logic is a logic that measures utility (as a real value in the interval [0,1]) instead of discrete truth (only 0 or 1). It is able to express properties that give more weight to the near future than to the far future. This article extends earlier work on DCTL with time, to continuous-time Markov chains and continuoustime Markov decision processes. It presents model checking algorithms for the two possible semantics of DCTL. This technical report is an extended version of [4] and is available from http://www.cs. ru.nl/research/reports.
منابع مشابه
Solving Continuous-Time Transition-Independent DEC-MDP with Temporal Constraints
Despite the impact of DEC-MDPs over the past decade, scaling to large problem domains has been difficult to achieve. The scale-up problem is exacerbated in DEC-MDPs with continuous states, which are critical in domains involving time; the latest algorithm (M-DPFP) does not scale-up beyond two agents and a handful of unordered tasks per agent. This paper is focused on meeting this challenge in c...
متن کاملA Fast Analytical Algorithm for Solving Markov Decision Processes with Real-Valued Resources
Agents often have to construct plans that obey deadlines or, more generally, resource limits for real-valued resources whose consumption can only be characterized by probability distributions, such as execution time or battery power. These planning problems can be modeled with continuous state Markov decision processes (MDPs) but existing solution methods are either inefficient or provide no gu...
متن کاملReduction of Discounted Continuous-Time MDPs with Unbounded Jump and Reward Rates to Discrete-Time Total-Reward MDPs
This article discusses a reduction of discounted Continuous-Time Markov Decision Processes (CTMDPs) to discrete-time Markov Decision Processes (MDPs). This reduction is based on the equivalence of a randomized policy that chooses actions only at jump epochs to a nonrandomized policy that can switch actions between jumps. For discounted CTMDPs with bounded jump rates, this reduction was introduc...
متن کاملPolicy gradients in linearly-solvable MDPs
We present policy gradient results within the framework of linearly-solvable MDPs. For the first time, compatible function approximators and natural policy gradients are obtained by estimating the cost-to-go function, rather than the (much larger) state-action advantage function as is necessary in traditional MDPs. We also develop the first compatible function approximators and natural policy g...
متن کاملBATHTUB HAZARD RATE DISTRIBUTIONS AND FUZZY LIFE TIMES
The development of life time analysis started back in the $20^{textit{th}}$ century and since then comprehensive developments have been made to model life time data efficiently. Recent development in measurements shows that all continuous measurements can not be measured as precise numbers but they are more or less fuzzy. Life time is also a continuous phenomenon, and has already been shown tha...
متن کامل